Lossy Compression of Noisy Data for Private and Data-Efficient Learning

نویسندگان

چکیده

Storage-efficient privacy-preserving learning is crucial due to increasing amounts of sensitive user data required for modern tasks. We propose a framework reducing the storage cost while at same time providing privacy guarantees, without essential loss in utility learning. Our method comprises noise injection followed by lossy compression. show that, when appropriately matching compression distribution added noise, compressed examples converge, distribution, that noise-free training as sample size (or dimension data) increases. In this sense, essentially maintained, and leakage quantifiable amounts. present experimental results on CelebA dataset gender classification find our suggested pipeline delivers practice promise theory: individuals images are unrecognizable less recognizable, depending level), overall substantially reduced, with no (and some cases slight boost) accuracy. As an bonus, experiments suggest yields substantial boost robustness face adversarial test data.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Lossless and Lossy Data Compression

Data compression (or source coding) is the process of creating binary representations of data which require less storage space than the original data 7; 14; 15]. Lossless compression is used where perfect reproduction is required while lossy compression is used where perfect reproduction is not possible or requires too many bits. Achieving optimal compression with respect to resource constraint...

متن کامل

Lossy Compression of Elevation Data

Generic lossy image compression algorithms, such as Said & Pearlman's, perform excellently on gridded terrain elevation data. For example, an average DEM compressed down to 0.1 bits per point has an RMS elevation error of 3 meters, although di erent DEMs ranged from 0.25 to 15 meters. This performance compares favorably to compressing with Triangulated Irregular Networks. Preliminary tests also...

متن کامل

Lossy Compression of Turbulence Data

A number of Grand Challenge applications are capable of generating such large quantities of data that data output may affect the applications performance. An example of such an application is modeling of turbulent convection. One means of dealing with this situation is data compression. This thesis examines applying wavelet based lossy compression techniques to turbulent convection data. It als...

متن کامل

Lossy Data Compression Using Logarithm

Lossy compression algorithms take advantage of the inherent limitations of the human eye and discard information that cannot be seen. [1] In the present paper a technique termed as Lossy Data Compression using Logarithm (LDCL) is proposed to compress incoming binary data in the form of a resultant matrix containing the logarithmic values of different chosen numeric sets. The proposed method is ...

متن کامل

Pointwise redundancy in lossy data compression and universal lossy data compression

We characterize the achievable pointwise redundancy rates for lossy data compression at a fixed distortion level. “Pointwise redundancy” refers to the difference between the description length achieved by an nth-order block code and the optimal nR(D) bits. For memoryless sources, we show that the best achievable redundancy rate is of order O( √ n) in probability. This follows from a second-orde...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE journal on selected areas in information theory

سال: 2022

ISSN: ['2641-8770']

DOI: https://doi.org/10.1109/jsait.2023.3260720